Orthogonal Tensor Decomposition
نویسنده
چکیده
In symmetric tensor decomposition one expresses a given symmetric tensor T a sum of tensor powers of a number of vectors: T = v⊗d 1 + · · · + v ⊗d k . Orthogonal decomposition is a special type of symmetric tensor decomposition in which in addition the vectors v1, ..., vk are required to be pairwise orthogonal. We study the properties of orthogonally decomposable tensors. In particular, we give a formula for all of the eigenvectors of an orthogonally decomposable tensor. Moreover, we give a conjecture for the defining equations of the set of orthogonally decomposable tensors.
منابع مشابه
Robust polynomial time tensor decomposition
Tensor decomposition has recently become an invaluable algorithmic primitive. It has seen much use in new algorithms with provable guarantees for fundamental statistics and machine learning problems. In these settings, some low-rank k-tensor A ∑r i 1 a ⊗k i which wewould like to decompose into components a1, . . . , ar ∈ n is often not directly accessible. This could happen for many reasons; a...
متن کاملSymmetric Orthogonal Tensor Decomposition is Trivial
We consider the problem of decomposing a real-valued symmetric tensor as the sum of outer products of real-valued, pairwise orthogonal vectors. Such decompositions do not generally exist, but we show that some symmetric tensor decomposition problems can be converted to orthogonal problems following the whitening procedure proposed by Anandkumar et al. (2012). If an orthogonal decomposition of a...
متن کاملA Counterexample to the Possibility of an Extension of the Eckart-Young Low-Rank Approximation Theorem for the Orthogonal Rank Tensor Decomposition
Earlier work has shown that no extension of the Eckart–Young SVD approximation theorem can be made to the strong orthogonal rank tensor decomposition. Here, we present a counterexample to the extension of the Eckart–Young SVD approximation theorem to the orthogonal rank tensor decomposition, answering an open question previously posed by Kolda [SIAM J. Matrix Anal. Appl., 23 (2001), pp. 243–355].
متن کاملOn the Tensor Svd and Optimal Low Rank Orthogonal Approximations of Tensors
Abstract. It is known that a high order tensor does not necessarily have an optimal low rank approximation, and that a tensor might not be orthogonally decomposable (i.e., admit a tensor SVD). We provide several sufficient conditions which lead to the failure of the tensor SVD, and characterize the existence of the tensor SVD with respect to the Higher Order SVD (HOSVD) of a tensor. In face of ...
متن کاملNon-orthogonal tensor diagonalization
Tensor diagonalization means transforming a given tensor to an exactly or nearly diagonal form through multiplying the tensor by non-orthogonal invertible matrices along selected dimensions of the tensor. It has a link to an approximate joint diagonalization (AJD) of a set of matrices. In this paper, we derive (1) a new algorithm for a symmetric AJD, which is called two-sided symmetric diagonal...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014